From b1d398b67781140d1c6efd05778d0ad4103b2a32 Mon Sep 17 00:00:00 2001 From: =?utf8?q?Roger=20Pau=20Monn=C3=A9?= Date: Thu, 26 Nov 2015 16:01:27 +0100 Subject: [PATCH] x86: allow disabling the emulated local apic MIME-Version: 1.0 Content-Type: text/plain; charset=utf8 Content-Transfer-Encoding: 8bit Signed-off-by: Roger Pau Monné Reviewed-by: Andrew Cooper Acked-by: Jan Beulich Acked-by: Kevin Tian --- xen/arch/x86/hvm/vlapic.c | 27 ++++++++++++++++++++++++--- xen/arch/x86/hvm/vmsi.c | 12 ++++++++++++ xen/arch/x86/hvm/vmx/vmcs.c | 5 ++++- xen/arch/x86/hvm/vmx/vmx.c | 6 ++++++ 4 files changed, 46 insertions(+), 4 deletions(-) diff --git a/xen/arch/x86/hvm/vlapic.c b/xen/arch/x86/hvm/vlapic.c index 15d59f652d..01a8430be5 100644 --- a/xen/arch/x86/hvm/vlapic.c +++ b/xen/arch/x86/hvm/vlapic.c @@ -988,6 +988,9 @@ static void set_x2apic_id(struct vlapic *vlapic) bool_t vlapic_msr_set(struct vlapic *vlapic, uint64_t value) { + if ( !has_vlapic(vlapic_domain(vlapic)) ) + return 0; + if ( (vlapic->hw.apic_base_msr ^ value) & MSR_IA32_APICBASE_ENABLE ) { if ( unlikely(value & MSR_IA32_APICBASE_EXTD) ) @@ -1200,6 +1203,9 @@ void vlapic_reset(struct vlapic *vlapic) struct vcpu *v = vlapic_vcpu(vlapic); int i; + if ( !has_vlapic(v->domain) ) + return; + vlapic_set_reg(vlapic, APIC_ID, (v->vcpu_id * 2) << 24); vlapic_set_reg(vlapic, APIC_LVR, VLAPIC_VERSION); @@ -1265,6 +1271,9 @@ static int lapic_save_hidden(struct domain *d, hvm_domain_context_t *h) struct vlapic *s; int rc = 0; + if ( !has_vlapic(d) ) + return 0; + for_each_vcpu ( d, v ) { s = vcpu_vlapic(v); @@ -1281,6 +1290,9 @@ static int lapic_save_regs(struct domain *d, hvm_domain_context_t *h) struct vlapic *s; int rc = 0; + if ( !has_vlapic(d) ) + return 0; + for_each_vcpu ( d, v ) { if ( hvm_funcs.sync_pir_to_irr ) @@ -1328,7 +1340,10 @@ static int lapic_load_hidden(struct domain *d, hvm_domain_context_t *h) uint16_t vcpuid; struct vcpu *v; struct vlapic *s; - + + if ( !has_vlapic(d) ) + return -ENODEV; + /* Which vlapic to load? */ vcpuid = hvm_load_instance(h); if ( vcpuid >= d->max_vcpus || (v = d->vcpu[vcpuid]) == NULL ) @@ -1360,7 +1375,10 @@ static int lapic_load_regs(struct domain *d, hvm_domain_context_t *h) uint16_t vcpuid; struct vcpu *v; struct vlapic *s; - + + if ( !has_vlapic(d) ) + return -ENODEV; + /* Which vlapic to load? */ vcpuid = hvm_load_instance(h); if ( vcpuid >= d->max_vcpus || (v = d->vcpu[vcpuid]) == NULL ) @@ -1399,7 +1417,7 @@ int vlapic_init(struct vcpu *v) HVM_DBG_LOG(DBG_LEVEL_VLAPIC, "%d", v->vcpu_id); - if ( is_pvh_vcpu(v) ) + if ( !has_vlapic(v->domain) ) { vlapic->hw.disabled = VLAPIC_HW_DISABLED; return 0; @@ -1452,6 +1470,9 @@ void vlapic_destroy(struct vcpu *v) { struct vlapic *vlapic = vcpu_vlapic(v); + if ( !has_vlapic(v->domain) ) + return; + tasklet_kill(&vlapic->init_sipi.tasklet); TRACE_0D(TRC_HVM_EMUL_LAPIC_STOP_TIMER); destroy_periodic_time(&vlapic->pt); diff --git a/xen/arch/x86/hvm/vmsi.c b/xen/arch/x86/hvm/vmsi.c index ac838a9a05..499151e28c 100644 --- a/xen/arch/x86/hvm/vmsi.c +++ b/xen/arch/x86/hvm/vmsi.c @@ -391,6 +391,9 @@ int msixtbl_pt_register(struct domain *d, struct pirq *pirq, uint64_t gtable) ASSERT(spin_is_locked(&pcidevs_lock)); ASSERT(spin_is_locked(&d->event_lock)); + if ( !has_vlapic(d) ) + return -ENODEV; + /* * xmalloc() with irq_disabled causes the failure of check_lock() * for xenpool->lock. So we allocate an entry beforehand. @@ -446,6 +449,9 @@ void msixtbl_pt_unregister(struct domain *d, struct pirq *pirq) ASSERT(spin_is_locked(&pcidevs_lock)); ASSERT(spin_is_locked(&d->event_lock)); + if ( !has_vlapic(d) ) + return; + irq_desc = pirq_spin_lock_irq_desc(pirq, NULL); if ( !irq_desc ) return; @@ -482,6 +488,9 @@ found: void msixtbl_init(struct domain *d) { + if ( !has_vlapic(d) ) + return; + INIT_LIST_HEAD(&d->arch.hvm_domain.msixtbl_list); spin_lock_init(&d->arch.hvm_domain.msixtbl_list_lock); @@ -493,6 +502,9 @@ void msixtbl_pt_cleanup(struct domain *d) struct msixtbl_entry *entry, *temp; unsigned long flags; + if ( !has_vlapic(d) ) + return; + /* msixtbl_list_lock must be acquired with irq_disabled for check_lock() */ local_irq_save(flags); spin_lock(&d->arch.hvm_domain.msixtbl_list_lock); diff --git a/xen/arch/x86/hvm/vmx/vmcs.c b/xen/arch/x86/hvm/vmx/vmcs.c index 000d06e387..7a7896e716 100644 --- a/xen/arch/x86/hvm/vmx/vmcs.c +++ b/xen/arch/x86/hvm/vmx/vmcs.c @@ -1035,7 +1035,7 @@ static int construct_vmcs(struct vcpu *v) ~(SECONDARY_EXEC_ENABLE_VM_FUNCTIONS | SECONDARY_EXEC_ENABLE_VIRT_EXCEPTIONS); - if ( is_pvh_domain(d) ) + if ( !has_vlapic(d) ) { /* Disable virtual apics, TPR */ v->arch.hvm_vmx.secondary_exec_control &= @@ -1047,7 +1047,10 @@ static int construct_vmcs(struct vcpu *v) /* In turn, disable posted interrupts. */ __vmwrite(PIN_BASED_VM_EXEC_CONTROL, vmx_pin_based_exec_control & ~PIN_BASED_POSTED_INTERRUPT); + } + if ( is_pvh_domain(d) ) + { /* Unrestricted guest (real mode for EPT) */ v->arch.hvm_vmx.secondary_exec_control &= ~SECONDARY_EXEC_UNRESTRICTED_GUEST; diff --git a/xen/arch/x86/hvm/vmx/vmx.c b/xen/arch/x86/hvm/vmx/vmx.c index 3eeccbb2c3..2581e97d52 100644 --- a/xen/arch/x86/hvm/vmx/vmx.c +++ b/xen/arch/x86/hvm/vmx/vmx.c @@ -89,6 +89,9 @@ static int vmx_domain_initialise(struct domain *d) { int rc; + if ( !has_vlapic(d) ) + return 0; + if ( (rc = vmx_alloc_vlapic_mapping(d)) != 0 ) return rc; @@ -97,6 +100,9 @@ static int vmx_domain_initialise(struct domain *d) static void vmx_domain_destroy(struct domain *d) { + if ( !has_vlapic(d) ) + return; + vmx_free_vlapic_mapping(d); } -- 2.30.2